lexical analysis - significado y definición. Qué es lexical analysis
Diclib.com
Diccionario ChatGPT
Ingrese una palabra o frase en cualquier idioma 👆
Idioma:

Traducción y análisis de palabras por inteligencia artificial ChatGPT

En esta página puede obtener un análisis detallado de una palabra o frase, producido utilizando la mejor tecnología de inteligencia artificial hasta la fecha:

  • cómo se usa la palabra
  • frecuencia de uso
  • se utiliza con más frecuencia en el habla oral o escrita
  • opciones de traducción
  • ejemplos de uso (varias frases con traducción)
  • etimología

Qué (quién) es lexical analysis - definición

COMPUTING PROCESS OF PARSING A SEQUENCE OF CHARACTERS INTO A SEQUENCE OF TOKENS
Lexical Analysis; Lexical analyzer; Token (parser); Lexer; Lexical token; Lexical analyser; Scanner (computing); Tokenize; Lexing; Tokenise; Tokenized; Tokenizing; Lexical parser; Tokenizer; Tokeniser; Tokenization (lexical analysis); Token splitting; Token scanner; Lexer generator; Lexer (computer science); Semicolon insertion; List of lexer generators; Lexical syntax; Lexeme (computer science); Automatic semicolon insertion; Lexers

lexical analysis         
<programming> (Or "linear analysis", "scanning") The first stage of processing a language. The stream of characters making up the source program or other input is read one at a time and grouped into lexemes (or "tokens") - word-like pieces such as keywords, identifiers, literals and punctutation. The lexemes are then passed to the parser. ["Compilers - Principles, Techniques and Tools", by Alfred V. Aho, Ravi Sethi and Jeffrey D. Ullman, pp. 4-5] (1995-04-05)
Lexical analysis         
In computer science, lexical analysis, lexing or tokenization is the process of converting a sequence of characters (such as in a computer program or web page) into a sequence of lexical tokens (strings with an assigned and thus identified meaning). A program that performs lexical analysis may be termed a lexer, tokenizer, or scanner, although scanner is also a term for the first stage of a lexer.
lexical analyser         
<language> (Or "scanner") The initial input stage of a language processor (e.g. a compiler), the part that performs lexical analysis. (1995-04-05)

Wikipedia

Lexical analysis

In computer science, lexical analysis, lexing or tokenization is the process of converting a sequence of characters (such as in a computer program or web page) into a sequence of lexical tokens (strings with an assigned and thus identified meaning). A program that performs lexical analysis may be termed a lexer, tokenizer, or scanner, although scanner is also a term for the first stage of a lexer. A lexer is generally combined with a parser, which together analyze the syntax of programming languages, web pages, and so forth.